Efficient kernel-based variable selection with sparsistency

نویسندگان

چکیده

Variable selection is central to high-dimensional data analysis, and various algorithms have been developed. Ideally, a variable algorithm shall be flexible, scalable, with theoretical guarantee, yet most existing cannot attain these properties at the same time. In this article, three-step developed, involving kernel-based estimation of regression function its gradient functions as well hard thresholding. Its key advantage that it assumes no explicit model assumption, admits general predictor effects, allows for scalable computation, attains desirable asymptotic sparsistency. The proposed can adapted any reproducing kernel Hilbert space (RKHS) different functions, extended interaction slight modification. computational cost only linear in dimension, further improved through parallel computing. sparsistency established RKHS under mild conditions, including Gaussian kernels special cases. effectiveness also supported by variety simulated real examples.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Scalable kernel-based variable selection with sparsistency

Variable selection is central to high-dimensional data analysis, and various algorithms have been developed. Ideally, a variable selection algorithm shall be flexible, scalable, and with theoretical guarantee, yet most existing algorithms cannot attain these properties at the same time. In this article, a three-step variable selection algorithm is developed, involving kernel-based estimation of...

متن کامل

Bayesian Approximate Kernel Regression with Variable Selection

Nonlinear kernel regression models are often used in statistics and machine learning due to greater accuracy than linear models. Variable selection for kernel regression models is a challenge partly because, unlike the linear regression setting, there is no clear concept of an effect size for regression coefficients. In this paper, we propose a novel framework that provides an analog of the eff...

متن کامل

Scalable Bayesian Kernel Models with Variable Selection

Nonlinear kernels are used extensively in regression models in statistics and machine learning since they often improve predictive accuracy. Variable selection is a challenge in the context of kernel based regression models. In linear regression the concept of an effect size for the regression coefficients is very useful for variable selection. In this paper we provide an analog for the effect ...

متن کامل

Gradient-based kernel method for feature extraction and variable selection

We propose a novel kernel approach to dimension reduction for supervised learning: feature extraction and variable selection; the former constructs a small number of features from predictors, and the latter finds a subset of predictors. First, a method of linear feature extraction is proposed using the gradient of regression function, based on the recent development of the kernel method. In com...

متن کامل

Efficient computer experiment-based optimization through variable selection

A computer experiment-based optimization approach employs design of experiments and statistical modeling to represent a complex objective function that can only be evaluated pointwise by running a computer model. In large-scale applications, the number of variables is huge, and direct use of computer experiments would require an exceedingly large experimental design and, consequently, significa...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Statistica Sinica

سال: 2021

ISSN: ['1017-0405', '1996-8507']

DOI: https://doi.org/10.5705/ss.202019.0401